Consistent Selection of Tuning Parameters via Variable Selection Stability ( Supplementary Material )

نویسندگان

  • Wei Sun
  • Junhui Wang
  • Yixin Fang
  • Xiaotong Shen
چکیده

In this supplementary material, we provide Lemmas 2 and 3 and their proofs. Suppose that x 1 ,. .. , x n are i.i.d. from a probability distribution with mean 0 and finite covariance matrix C = (C jk). Assumption S1: Assume that x 1 has finite fourth moment, that is, E(x 1i x 1 j x 1k x 1l) is finite for any 1 ≤ i, j, k, l ≤ p. Lemma 2 Suppose that Assumption S1 is met. Assumptions 1 and 2 are satisfied by the lasso regression and the SCAD with r n = n −1/2 and s n = o(1) under the assumptions in Zhao and Yu (2006) or Fan and Li (2001), and by the adaptive lasso with r n = n −1 and s n = n −1/2 under the assumptions in Zou (2006), on the random splitting subsamples generated in Algorithm 1.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Consistent selection of tuning parameters via variable selection stability

Penalized regression models are popularly used in high-dimensional data analysis to conduct variable selection and model fitting simultaneously. Whereas success has been widely reported in literature, their performances largely depend on the tuning parameters that balance the trade-off between model fitting and model sparsity. Existing tuning criteria mainly follow the route of minimizing the e...

متن کامل

Tuning Variable Selection Procedures by Adding Noise

Many variable selection methods for linear regression depend critically on tuning parameters that control the performance of the method, e.g., “entry” and“stay” significance levels in forward and backward selection. However, most methods do not adapt the tuning parameters to particular data sets. We propose a general strategy for adapting variable selection tuning parameters that effectively es...

متن کامل

A Monte Carlo-Based Search Strategy for Dimensionality Reduction in Performance Tuning Parameters

Redundant and irrelevant features in high dimensional data increase the complexity in underlying mathematical models. It is necessary to conduct pre-processing steps that search for the most relevant features in order to reduce the dimensionality of the data. This study made use of a meta-heuristic search approach which uses lightweight random simulations to balance between the exploitation of ...

متن کامل

Stability Selection for Structured Variable Selection

In variable or graph selection problems, finding a right-sized model or controlling the number of false positives is notoriously difficult. Recently, a meta-algorithm called Stability Selection was proposed that can provide reliable finite-sample control of the number of false positives. Its benefits were demonstrated when used in conjunction with the lasso and orthogonal matching pursuit algor...

متن کامل

Variable Selection from Random Forests: Application to Gene Expression Data

Random forest is a classification algorithm well suited for microarray data: it shows excellent performance even when most predictive variables are noise, can be used when the number of variables is much larger than the number of observations, and returns measures of variable importance. Thus, it is important to understand the performance of random forest with microarray data and its use for ge...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013